- Home
- Search Results
- Page 1 of 1
Search for: All records
-
Total Resources4
- Resource Type
-
0003010000000000
- More
- Availability
-
40
- Author / Contributor
- Filter by Author / Creator
-
-
Darrell, Trevor (4)
-
Rohrbach, Anna (4)
-
Gonzalez, Joseph E (2)
-
Petryk, Suzanne (2)
-
Rohrbach, Marcus (2)
-
Whitehead, Spencer (2)
-
Dunlap, Lisa (1)
-
Gardner, Matt (1)
-
Gonzalez, Joseph_E (1)
-
Guillory, Devin (1)
-
Merrill, William (1)
-
Mohri, Clara (1)
-
Raghunathan, Aditi (1)
-
Singh, Sameer (1)
-
Subramanian, Sanjay (1)
-
Zhang, Han (1)
-
#Tyler Phillips, Kenneth E. (0)
-
#Willis, Ciara (0)
-
& Abreu-Ramos, E. D. (0)
-
& Abramson, C. I. (0)
-
- Filter by Editor
-
-
& Spizer, S. M. (0)
-
& . Spizer, S. (0)
-
& Ahn, J. (0)
-
& Bateiha, S. (0)
-
& Bosch, N. (0)
-
& Brennan K. (0)
-
& Brennan, K. (0)
-
& Chen, B. (0)
-
& Chen, Bodong (0)
-
& Drown, S. (0)
-
& Ferretti, F. (0)
-
& Higgins, A. (0)
-
& J. Peters (0)
-
& Kali, Y. (0)
-
& Ruiz-Arias, P.M. (0)
-
& S. Spitzer (0)
-
& Sahin. I. (0)
-
& Spitzer, S. (0)
-
& Spitzer, S.M. (0)
-
(submitted - in Review for IEEE ICASSP-2024) (0)
-
-
Have feedback or suggestions for a way to improve these results?
!
Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher.
Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?
Some links on this page may take you to non-federal websites. Their policies may differ from this site.
-
Petryk, Suzanne; Whitehead, Spencer; Gonzalez, Joseph_E; Darrell, Trevor; Rohrbach, Anna; Rohrbach, Marcus (, arXiv)
-
Dunlap, Lisa; Mohri, Clara; Guillory, Devin; Zhang, Han; Darrell, Trevor; Gonzalez, Joseph E; Raghunathan, Aditi; Rohrbach, Anna (, International Conference on Learning Representations (ICLR))
-
Subramanian, Sanjay; Merrill, William; Darrell, Trevor; Gardner, Matt; Singh, Sameer; Rohrbach, Anna (, Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers))Training a referring expression comprehension (ReC) model for a new visual domain requires collecting referring expressions, and potentially corresponding bounding boxes, for images in the domain. While large-scale pre-trained models are useful for image classification across domains, it remains unclear if they can be applied in a zero-shot manner to more complex tasks like ReC. We present ReCLIP, a simple but strong zero-shot baseline that repurposes CLIP, a state-of-the-art large-scale model, for ReC. Motivated by the close connection between ReC and CLIP’s contrastive pre-training objective, the first component of ReCLIP is a region-scoring method that isolates object proposals via cropping and blurring, and passes them to CLIP. However, through controlled experiments on a synthetic dataset, we find that CLIP is largely incapable of performing spatial reasoning off-the-shelf. We reduce the gap between zero-shot baselines from prior work and supervised models by as much as 29% on RefCOCOg, and on RefGTA (video game imagery), ReCLIP’s relative improvement over supervised ReC models trained on real images is 8%.more » « less
An official website of the United States government

Full Text Available